pricing strategy
Online Dynamic Pricing of Complementary Products
Mussi, Marco, Restelli, Marcello
Traditional pricing paradigms, once dominated by static models and rule-based heuristics, are increasingly being replaced by dynamic, data-driven approaches powered by machine learning algorithms. Despite their growing sophistication, most dynamic pricing algorithms focus on optimizing the price of each product independently, disregarding potential interactions among items. By neglecting these interdependencies in consumer demand across related goods, sellers may fail to capture the full potential of coordinated pricing strategies. In this paper, we address this problem by exploring dynamic pricing mechanisms designed explicitly for complementary products, aiming to exploit their joint demand structure to maximize overall revenue. We present an online learning algorithm considering both positive and negative interactions between products' demands. The algorithm utilizes transaction data to identify advantageous complementary relationships through an integer programming problem between different items, and then optimizes pricing strategies using data-driven and computationally efficient multi-armed bandit solutions based on heteroscedastic Gaussian processes. We validate our solution in a simulated environment, and we demonstrate that our solution improves the revenue w.r.t. a comparable learning algorithm ignoring such interactions.
- Research Report (1.00)
- Overview (1.00)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Africa > Central African Republic > Ombella-M'Poko > Bimbo (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- (2 more...)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Africa > Central African Republic > Ombella-M'Poko > Bimbo (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- (2 more...)
Integrating Attention-Enhanced LSTM and Particle Swarm Optimization for Dynamic Pricing and Replenishment Strategies in Fresh Food Supermarkets
Liu, Xianchen, Zhang, Tianhui, Zhang, Xinyu, Hou, Lingmin, Guo, Zhen, Tian, Yuanhao, Liu, Yang
This paper presents a novel approach to optimizing pricing and replenishment strategies in fresh food supermarkets by combining Long Short-Term Memory (LSTM) networks with Particle Swarm Optimization (PSO). The LSTM model, enhanced with an attention mechanism, is used to predict sales volumes, pricing trends, and spoilage rates over a seven-day period. The predictions generated by the LSTM model serve as inputs for the PSO algorithm, which iteratively optimizes pricing and replenishment strategies to maximize profitability while adhering to inventory constraints. The integration of cost-plus pricing allows for dynamic adjustments based on fixed and variable costs, ensuring real-time adaptability to market fluctuations. The framework not only maximizes profits but also reduces food waste, contributing to more sustainable supermarket operations. The attention mechanism enhances the interpretability of the LSTM model by identifying key time points and factors influencing sales, improving decision-making accuracy. This methodology bridges the gap between predictive modeling and optimization, offering a scalable solution for dynamic pricing and inventory management in fresh food retail and other industries dealing with perishable goods.
- North America > United States > Florida > Miami-Dade County > Miami (0.05)
- Africa > Nigeria > Lagos State > Lagos (0.04)
- North America > United States > Massachusetts > Suffolk County > Boston (0.04)
- North America > Trinidad and Tobago > Trinidad > Arima > Arima (0.04)
- Overview > Innovation (0.66)
- Research Report > Promising Solution (0.48)
- Retail (1.00)
- Consumer Products & Services > Food, Beverage, Tobacco & Cannabis (1.00)
- Banking & Finance (1.00)
What is in a Price? Estimating Willingness-to-Pay with Bayesian Hierarchical Models
Pillai, Srijesh, Chandrawat, Rajesh Kumar
For premium consumer products, pricing strategy is not about a single number, but about understanding the perceived monetary value of the features that justify a higher cost. This paper proposes a robust methodology to deconstruct a product's price into the tangible value of its constituent parts. We employ Bayesian Hierarchical Conjoint Analysis, a sophisticated statistical technique, to solve this high-stakes business problem using the Apple iPhone as a universally recognizable case study. We first simulate a realistic choice based conjoint survey where consumers choose between different hypothetical iPhone configurations. We then develop a Bayesian Hierarchical Logit Model to infer consumer preferences from this choice data. The core innovation of our model is its ability to directly estimate the Willingness-to-Pay (WTP) in dollars for specific feature upgrades, such as a "Pro" camera system or increased storage. Our results demonstrate that the model successfully recovers the true, underlying feature valuations from noisy data, providing not just a point estimate but a full posterior probability distribution for the dollar value of each feature. This work provides a powerful, practical framework for data-driven product design and pricing strategy, enabling businesses to make more intelligent decisions about which features to build and how to price them.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Asia > Middle East > UAE > Dubai Emirate > Dubai (0.04)
- North America > United States > New York (0.04)
- Europe > Netherlands (0.04)
Multi-Agent Reinforcement Learning for Dynamic Pricing in Supply Chains: Benchmarking Strategic Agent Behaviours under Realistically Simulated Market Conditions
Hazenberg, Thomas, Ma, Yao, Ziabari, Seyed Sahand Mohammadi, van Rijswijk, Marijn
This study investigates how Multi-Agent Reinforcement Learning (MARL) can improve dynamic pricing strategies in supply chains, particularly in contexts where traditional ERP systems rely on static, rule-based approaches that overlook strategic interactions among market actors. While recent research has applied reinforcement learning to pricing, most implementations remain single-agent and fail to model the interdependent nature of real-world supply chains. This study addresses that gap by evaluating the performance of three MARL algorithms: MADDPG, MADQN, and QMIX against static rule-based baselines, within a simulated environment informed by real e-commerce transaction data and a LightGBM demand prediction model. Results show that rule-based agents achieve near-perfect fairness (Jain's Index: 0.9896) and the highest price stability (volatility: 0.024), but they fully lack competitive dynamics. Among MARL agents, MADQN exhibits the most aggressive pricing behaviour, with the highest volatility and the lowest fairness (0.5844). MADDPG provides a more balanced approach, supporting market competition (share volatility: 9.5 pp) while maintaining relatively high fairness (0.8819) and stable pricing. These findings suggest that MARL introduces emergent strategic behaviour not captured by static pricing rules and may inform future developments in dynamic pricing.
- Europe > United Kingdom (0.14)
- Asia > China > Heilongjiang Province > Daqing (0.04)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- (7 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.93)
Learning Personalized Utility Functions for Drivers in Ride-hailing Systems Using Ensemble Hypernetworks
Mai, Weiming, Gao, Jie, Cats, Oded
In ride-hailing systems, drivers decide whether to accept or reject ride requests based on factors such as order characteristics, traffic conditions, and personal preferences. Accurately predicting these decisions is essential for improving the efficiency and reliability of these systems. Traditional models, such as the Random Utility Maximization (RUM) approach, typically predict drivers' decisions by assuming linear correlations among attributes. However, these models often fall short because they fail to account for non-linear interactions between attributes and do not cater to the unique, personalized preferences of individual drivers. In this paper, we develop a method for learning personalized utility functions using hypernetwork and ensemble learning. Hypernetworks dynamically generate weights for a linear utility function based on trip request data and driver profiles, capturing the non-linear relationships. An ensemble of hypernetworks trained on different data segments further improve model adaptability and generalization by introducing controlled randomness, thereby reducing over-fitting. We validate the performance of our ensemble hypernetworks model in terms of prediction accuracy and uncertainty estimation in a real-world dataset. The results demonstrate that our approach not only accurately predicts each driver's utility but also effectively balances the needs for explainability and uncertainty quantification. Additionally, our model serves as a powerful tool for revealing the personalized preferences of different drivers, clearly illustrating which attributes largely impact their rider acceptance decisions.
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.96)
- Information Technology > Game Theory (0.86)
- Information Technology > Decision Support Systems (0.86)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.68)
Sam's Club is adding AI to the shopping experience. Why are privacy advocacy groups worried?
Sam's Club is going register-free and introducing an all-digital, AI-powered shopping experience for its customers, a move that has privacy advocates worried that the new AI tool could be used to unfairly target some customers with higher-priced items based on their shopping habits. The all-digital approach started with the reconstruction of a Sam's Club in Grapevine, a suburb of Dallas, that was severely damaged in 2022 by a tornado. When the retail location opened two years later it was the first of its kind to ditch its registers for a "Scan and Go" program that allowed customers to scan each item placed in their physical cart and pay through a mobile app. This program has since been piloted in nine Dallas metro locations and one store in Missouri, Retail Dive reported. Instead of handing a receipt to a Sam's Club employee to review before leaving the store, customers walk through an arch that's equipped with AI-powered cameras to capture images of the items in the cart and electronically match them with the items paid for through the app. Sam's Club did not disclose when the AI technology would be coming to California stores but Sam's Club has outlets in Torrance, Fountain Valley, El Monte and Riverside.
- North America > United States > California (0.62)
- North America > United States > Missouri (0.25)
- Retail (1.00)
- Information Technology > Security & Privacy (0.36)
Dynamic Assortment Selection and Pricing with Censored Preference Feedback
In this study, we investigate the problem of dynamic multi-product selection and pricing by introducing a novel framework based on a \textit{censored multinomial logit} (C-MNL) choice model. In this model, sellers present a set of products with prices, and buyers filter out products priced above their valuation, purchasing at most one product from the remaining options based on their preferences. The goal is to maximize seller revenue by dynamically adjusting product offerings and prices, while learning both product valuations and buyer preferences through purchase feedback. To achieve this, we propose a Lower Confidence Bound (LCB) pricing strategy. By combining this pricing strategy with either an Upper Confidence Bound (UCB) or Thompson Sampling (TS) product selection approach, our algorithms achieve regret bounds of $\tilde{O}(d^{\frac{3}{2}}\sqrt{T/\kappa})$ and $\tilde{O}(d^{2}\sqrt{T/\kappa})$, respectively. Finally, we validate the performance of our methods through simulations, demonstrating their effectiveness.
- Asia > South Korea > Seoul > Seoul (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
Learn then Decide: A Learning Approach for Designing Data Marketplaces
Gao, Yingqi, Zhou, Jin, Zhou, Hua, Chen, Yong, Dai, Xiaowu
As data marketplaces become increasingly central to the digital economy, it is crucial to design efficient pricing mechanisms that optimize revenue while ensuring fair and adaptive pricing. We introduce the Maximum Auction-to-Posted Price (MAPP) mechanism, a novel two-stage approach that first estimates the bidders' value distribution through auctions and then determines the optimal posted price based on the learned distribution. We establish that MAPP is individually rational and incentive-compatible, ensuring truthful bidding while balancing revenue maximization with minimal price discrimination. MAPP achieves a regret of $O_p(n^{-1})$ when incorporating historical bid data, where $n$ is the number of bids in the current round. It outperforms existing methods while imposing weaker distributional assumptions. For sequential dataset sales over $T$ rounds, we propose an online MAPP mechanism that dynamically adjusts pricing across datasets with varying value distributions. Our approach achieves no-regret learning, with the average cumulative regret converging at a rate of $O_p(T^{-1/2}(\log T)^2)$. We validate the effectiveness of MAPP through simulations and real-world data from the FCC AWS-3 spectrum auction.
- Asia > Middle East > Jordan (0.05)
- North America > United States > Oregon > Multnomah County > Portland (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)